A Hybrid Proactive Caching System in Vehicular Networks Based on Contextual Multi-Armed Bandit Learning

نویسندگان

چکیده

Proactive edge caching has been regarded as an effective approach to satisfy user experience in mobile networks by providing seamless content transmission and reducing network delay. This is particularly useful rapidly changing vehicular networks. paper addresses the proactive (at roadside unit (RSU)) problem mobility prediction, i.e., next RSU prediction. Specifically, proposes a distributed Hybrid cMAB Caching System where RSUs act independent learners that implement two parallel online reinforcement learning-based prediction algorithms between which they can adaptively finalize their predictions for RSU. The are based on xmlns:xlink="http://www.w3.org/1999/xlink">Contextual Multi-armed bandit (cMAB) learning, called xmlns:xlink="http://www.w3.org/1999/xlink">Dual-context cMAB xmlns:xlink="http://www.w3.org/1999/xlink">Single-context . hybrid system further developed into variants: xmlns:xlink="http://www.w3.org/1999/xlink">Vehicle-Centric xmlns:xlink="http://www.w3.org/1999/xlink">RSU-Centric In addition, also conducts comprehensive simulation experiments evaluate performance of proposed system. They include three traffic scenarios: xmlns:xlink="http://www.w3.org/1999/xlink">Commuting traffic, Random traffic xmlns:xlink="http://www.w3.org/1999/xlink">Mixed Las Vegas, USA Manchester, UK. With different road layouts urban areas, aims generalize application Simulation results show Vehicle-Centric reach nearly 95% cumulative accuracy Commuting scenario outperform other methods used comparison reaching 80% Mixed scenario. Even completely scenario, it guarantees minimum 60%.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Learning and decisions in contextual multi-armed bandit tasks

Contextual Multi-Armed Bandit (CMAB) tasks are a novel framework to assess decision making in uncertain environments. In a CMAB task, participants are presented with multiple options (arms) which are characterized by a number of features (context) related to the reward associated with the arms. By choosing arms repeatedly and observing the reward, participants can learn about the relation betwe...

متن کامل

Stream-based Online Active Learning in a Contextual Multi-Armed Bandit Framework

We study the stream-based online active learning in a contextual multi-armed bandit framework. In this framework, the reward depends on both the arm and the context. In a stream-based active learning setting, obtaining the ground truth of the reward is costly, and the conventional contextual multi-armed bandit algorithm fails to achieve a sublinear regret due to this cost. Hence, the algorithm ...

متن کامل

Human behavior in contextual multi-armed bandit problems

In real-life decision environments people learn from their direct experience with alternative courses of action. Yet they can accelerate their learning by using functional knowledge about the features characterizing the alternatives. We designed a novel contextual multi-armed bandit task where decision makers chose repeatedly between multiple alternatives characterized by two informative featur...

متن کامل

MULTI–ARMED BANDIT FOR PRICING Multi–Armed Bandit for Pricing

This paper is about the study of Multi–Armed Bandit (MAB) approaches for pricing applications, where a seller needs to identify the selling price for a particular kind of item that maximizes her/his profit without knowing the buyer demand. We propose modifications to the popular Upper Confidence Bound (UCB) bandit algorithm exploiting two peculiarities of pricing applications: 1) as the selling...

متن کامل

Variational inference for the multi-armed contextual bandit

In many biomedical, science, and engineering problems, one must sequentially decide which action to take next so as to maximize rewards. Reinforcement learning is an area of machine learning that studies how this maximization balances exploration and exploitation, optimizing interactions with the world while simultaneously learning how the world operates. One general class of algorithms for thi...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE Access

سال: 2023

ISSN: ['2169-3536']

DOI: https://doi.org/10.1109/access.2023.3259547